Fast and Scalable Structural SVM with Slack Rescaling

نویسندگان

  • Heejin Choi
  • Ofer Meshi
  • Nathan Srebro
چکیده

We present an efficient method for training slackrescaled structural SVM. Although finding the most violating label in a margin-rescaled formulation is often easy since the target function decomposes with respect to the structure, this is not the case for a slack-rescaled formulation, and finding the most violated label might be very difficult. Our core contribution is an efficient method for finding the most-violatinglabel in a slack-rescaled formulation, given an oracle that returns the most-violating-label in a (slightly modified) margin-rescaled formulation. We show that our method enables accurate and scalable training for slack-rescaled SVMs, reducing runtime by an order of magnitude compared to previous approaches to slack-rescaled SVMs.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Slack and Margin Rescaling as Convex Extensions of Supermodular Functions

Slack and margin rescaling are variants of the structured output SVM. They define convex surrogates to task specific loss functions, which, when specialized to nonadditive loss functions for multi-label problems, yield extensions to increasing set functions. We demonstrate in this paper that we may use these concepts to define polynomial time convex extensions of arbitrary supermodular function...

متن کامل

Supplementary Material for “DivMCuts: Faster Training of Structural SVMs with Diverse M-Best Cutting-Planes”

This section reviews the notation used in the main paper and revisits cutting-plane methods for training structured-output predictors. Notation. For any positive integer n we use [n] as shorthand for the set {1, 2, . . . , n}. We use y for a structured-output, and Y = (y1, . . . ,y|Y|) for a tuple of structured-outputs. Given a training dataset of input-output pairs {(xi,yi) |xi ∈X ,yi ∈Y}, we ...

متن کامل

A Parallel Way to Select the Parameters of SVM Based on the Ant Optimization Algorithm

A large number of experimental data shows that Support Vector Machine (SVM) algorithm has obvious a large advantages in text classification, handwriting recognition, image classification, bioinformatics, and some other fields. To some degree, the optimization of SVM depends on its kernel function and Slack variable, the determinant of which is its parameters δ and c in the classification functi...

متن کامل

A Note on the Generalization Performance of Kernel Classifiers with Margin

We present distribution independent bounds on the generalization misclassification performance of a family of kernel classifiers with margin. Support Vector Machine classifiers (SVM) stem out of this class of machines. The bounds are derived through computations of the Vγ dimension of a family of loss functions where the SVM one belongs to. Bounds that use functions of margin distributions (i.e...

متن کامل

The Benefits of Modeling Slack Variables in SVMs

In this letter, we explore the idea of modeling slack variables in support vector machine (SVM) approaches. The study is motivated by SVM+, which models the slacks through a smooth correcting function that is determined by additional (privileged) information about the training examples not available in the test phase. We take a closer look at the meaning and consequences of smooth modeling of s...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2016